Cross-modal description of sentiment information embedded in speech

نویسندگان

  • Kanako Watanabe
  • Yoko Greenberg
  • Yoshinori Sagisaka
چکیده

Looking for new possibilities to describe the information embedded in speech, we have carried out sentiment correlation analysis between speech features and color attributes. Using single vowel utterances with different prosody and sound pressure level, we have asked subjects to select colors based on their perceptual impressions after listening them. By analyzing selected color attributes using Value, Saturation and Hue, we found high correlations between mean F0 and Value, sound pressure level and Saturation, and Formants and Hue. These correlations coincided with previous observations using speech and color categories, which suggests a possibility for visualization of sentiment information embedded in speech based on cross-modal sentiment correlations.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Cross-Modal Analysis Between Phonation Differences and Texture Images Based on Sentiment Correlations

Motivated by the success of speech characteristics representation by color attributes, we analyzed the cross-modal sentiment correlations between voice source characteristics and textural image characteristics. For the analysis, we employed vowel sounds with representative three phonation differences (modal, creaky and breathy) and 36 texture images with 36 semantic attributes (e.g., banded, cr...

متن کامل

YouTube Movie Reviews: In, Cross, and Open-domain Sentiment Analysis in an Audiovisual Context

In this contribution we focus on the task of automatically analyzing a speaker’s sentiment in on-line videos containing movie reviews. In addition to textual information, we consider adding audio features as typically used in speech-based emotion recognition as well as video features encoding valuable valence information conveyed by the speaker. We combine this multi-modal experimental setup wi...

متن کامل

Cross-modal Association between Auditory and Visuospatial Information in Mandarin Tone Perception in Noise by Native and Non-native Perceivers

Speech perception involves multiple input modalities. Research has indicated that perceivers establish cross-modal associations between auditory and visuospatial events to aid perception. Such intermodal relations can be particularly beneficial for speech development and learning, where infants and non-native perceivers need additional resources to acquire and process new sounds. This study exa...

متن کامل

Lexical Segmentation and Ambiguity: Investigating the Recognition of Onset-embedded Words

The lack of acoustic markers of word boundaries in connected speech may create temporary ambiguities between words like cap and the start of longer words like captain. These ambiguities have motivated models of spoken word recognition in which lexical competition allows information after the end of an embedded word to assist identification. We review the results of priming experiments demonstra...

متن کامل

Leading Up the Lexical Garden Path: Segmentation and Ambiguity in Spoken Word Recognition

Two gating studies, a forced-choice identification study and 2 series of cross-modal repetition priming experiments, traced the time course of recognition of words with onset embeddings (captain) and short words in contexts that match (cap tucked) or mismatch (cap looking) with longer words. Results suggest that acoustic differences in embedded syllables assist the perceptual system in discrimi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2015